Scalable Graph Convolutional Network Training on Distributed-Memory Systems

نویسندگان

چکیده

Graph Convolutional Networks (GCNs) are extensively utilized for deep learning on graphs. The large data sizes of graphs and their vertex features make scalable training algorithms distributed memory systems necessary. Since the convolution operation induces irregular access patterns, designing a memory- communication-efficient parallel algorithm GCN poses unique challenges. We propose highly that scales to processor counts. In our solution, adjacency vertex-feature matrices partitioned among processors. exploit vertex-partitioning graph use non-blocking point-to-point communication operations between processors better scalability. To further minimize parallelization overheads, we introduce sparse matrix partitioning scheme based hypergraph model full-batch training. also novel stochastic encode expected volume in mini-batch show merits model, previously unexplored training, over standard which does not accurately costs. Experiments performed real-world datasets demonstrate proposed achieve considerable speedups alternative solutions. optimizations achieved costs become even more pronounced at high scalability with many performance benefits preserved deeper GCNs having layers as well billion-scale

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Learning Graph While Training: An Evolving Graph Convolutional Neural Network

Convolution Neural Networks on Graphs are important generalization and extension of classical CNNs. While previous works generally assumed that the graph structures of samples are regular with unified dimensions, in many applications, they are highly diverse or even not well defined. Under some circumstances, e.g. chemical molecular data, clustering or coarsening for simplifying the graphs is h...

متن کامل

Efficient Training of Convolutional Neural Nets on Large Distributed Systems

Deep Neural Networks (DNNs) have achieved impressive accuracy in many application domains including image classification. Training of DNNs is an extremely computeintensive process and is solved using variants of the stochastic gradient descent (SGD) algorithm. A lot of recent research has focussed on improving the performance of DNN training. In this paper, we present optimization techniques to...

متن کامل

A Scalable Parallel Graph Coloring Algorithm for Distributed Memory Computers

In large-scale parallel applications a graph coloring is often carried out to schedule computational tasks. In this paper, we describe a new distributedmemory algorithm for doing the coloring itself in parallel. The algorithm operates in an iterative fashion; in each round vertices are speculatively colored based on limited information, and then a set of incorrectly colored vertices, to be reco...

متن کامل

Distributed Graph Layout for Scalable Small-world Network Analysis

The in-memory graph layout or organization has a considerable impact on the time and energy efficiency of distributed memory graph computations. It affects memory locality, inter-task load balance, communication time, and overall memory utilization. Graph layout could refer to partitioning or replication of vertex and edge arrays, selective replication of data structures that hold meta-data, an...

متن کامل

Graph partitioning for scalable distributed graph computations

Inter-node communication time constitutes a significant fraction of the execution time of graph algorithms on distributed-memory systems. Global computations on large-scale sparse graphs with skewed degree distributions are particularly challenging to optimize for, as prior work shows that it is difficult to obtain balanced partitions with low edge cuts for these graphs. In this work, we attemp...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the VLDB Endowment

سال: 2022

ISSN: ['2150-8097']

DOI: https://doi.org/10.14778/3574245.3574256